Goto

Collaborating Authors

 generalized linear model








sup

Neural Information Processing Systems

This setup contains a vast array of fundamental applications in machine learning, engineering, neuroscience, finance, statisticsandinformation theory [1-10].


Learning with little mixing

Neural Information Processing Systems

We study square loss in a realizable time-series framework with martingale difference noise. Our main result is a fast rate excess risk bound which shows that whenever a trajectory hypercontractivity condition holds, the risk of the leastsquares estimator on dependent data matches the iid rate order-wise after a burn-in time. In comparison, many existing results in learning from dependent data have rates where the effective sample size is deflated by a factor of the mixing-time of the underlying process, even after the burn-in time. Furthermore, our results allow the covariate process to exhibit long range correlations which are substantially weaker than geometric ergodicity.



05b12f103c9e613efc4c85674cdc9066-Paper-Conference.pdf

Neural Information Processing Systems

Under label corruptions, we prove that this simple estimator achieves minimax near-optimal riskonawiderange ofgeneralized linear models, including Gaussian regression, Poisson regression and Binomial regression.